Efficient Cache Locking at Private First-Level Caches and Shared Last-Level Cache for Modern Multicore Systems
نویسندگان
چکیده
Most modern computing systems are having multicore processors with multilevel caches for high performance. Caches increase total power consumption and worsen execution time unpredictability. Studies show that way (or partial) cache locking may improve timing predictability and performance-to-power ratio for both single-core and multicore systems. Even though both private first-level and shared last-level cache locking improve timing predictability, it is difficult to justify the performance and power trade-off between these two locking mechanisms. In this work, we evaluate two cache locking schemes for multicore systems – at private first-level caches and at shared last-level cache. Both schemes are based on the analysis of applications’ worst case execution time (WCET) and both allow changing the locked cache size during runtime to achieve the optimal performance-to-power ratio for the running applications. Using Heptane WCET analyser, we generate workloads for H.264/AVC, MPEG4, FFT, MI, and DFT codes. Using VisualSim tool, we model and simulate a system with four cores and two levels of caches. Experimental results confirm that both cache locking schemes improve timing predictability by decreasing the total number of cache misses. Results also indicate that for small applications like FFT, shared last-level cache locking outperforms private first-level cache locking; but for large applications like MPEG4 and H.264/AVC, private first-level cache locking performs better than shared last-level cache locking. Keywords—Cache locking, multicore architecture, performance-to-power ratio, timing predictability, private first-level cache, shared last-cache
منابع مشابه
Static Task Partitioning for Locked Caches in Multi
Growing processing demand on multi-tasking real-time systems can be met by employing scalable multicore architectures. For such environments, locking cache lines for hard real-time systems ensures timing predictability of data references and may lower worst-case execution time. This work studies the benefits of cache locking on massive multicore architectures with private caches in the context ...
متن کاملUnderstanding Multicore Cache Behavior of Loop-based Parallel Programs via Reuse Distance Analysis
Understanding multicore memory behavior is crucial, but can be challenging due to the cache hierarchies employed in modern CPUs. In today’s hierarchies, performance is determined by complex thread interactions, such as interference in shared caches and replication and communication in private caches. Researchers normally perform simulation to sort out these interactions, but this can be costly ...
متن کاملA Power-Aware Multi-Level Cache Organization Effective for Multi-Core Embedded Systems
Recent system design trends suggest multicore architecture for all computing platforms including distributed and embedded systems running real-time applications. Multilevel caches in a multicore system pose serious challenges as cache requires huge amount of energy to be operated and cache increases unpredictability due to its dynamic behavior. Bandwidth and synchronization problems are also cr...
متن کاملA Daptive Block Pinning Based : D Ynamic C Ache Partitioning for M Ulti - Core Architectures
This paper is aimed at exploring the various techniques currently used for partitioning last level (L2/L3) caches in multicore architectures, identifying their strengths and weaknesses and thereby proposing a novel partitioning scheme known as Adaptive Block Pinning which would result in a better utilization of the cache resources in CMPs. The widening speed gap between processors and memory al...
متن کاملMANAGER: A Multicore Shared Cache Energy Saving Technique for QoS Systems
Last level caches (LLCs) contribute significantly to processor power consumption. Saving LLC energy in multicore QoS systems is especially challenging, since aggressive energy saving techniques may lead to failure in providing QoS. We present MANAGER, a multicore shared cache energy saving technique for quality-of-service systems. Using dynamic profiling, MANAGER periodically predicts cache acc...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
عنوان ژورنال:
دوره شماره
صفحات -
تاریخ انتشار 2015